Robots.txt Disallow command [on hold]
Posted
by
Saahil Sinha
on Pro Webmasters
See other posts from Pro Webmasters
or by Saahil Sinha
Published on 2013-06-23T02:47:18Z
Indexed on
2013/06/28
4:29 UTC
Read the original article
Hit count: 343
How to disallow folders through Robots.txt, which are been crawled due to wrong url structure, which thus cause duplicate page error
The URL been crawled as incorrectly by Google leading to duplicate page error:
www.abc.com/forum/index.php?option=com_forum
However, The actual correct pages however are:
www.abc.com/index.php?option=com_forum
Is this a correct way by excluding them through robots.txt: To exclude
www.abc.com/forum/index.php?option=com_forum
Below is command
Disallow: /forum/
Will it not block in legitimate component folder 'Forum' of site?
© Pro Webmasters or respective owner